Kullback–Leibler divergence

This captures the difference between the entropy of and the “wrong” entropy of measured with ().

A common way to measure “distance” between two distributions. It’s not a real distance because it’s not symmetric. The symmetrized version is called Jensen–Shannon divergence.

See also Gibbs’ inequality